Introspective Learning by Distilling Knowledge from Online Self-explanation

نویسندگان

چکیده

In recent years, many methods have been proposed to explain individual classification predictions of deep neural networks. However, how leverage the created explanations improve learning process has less explored. The extracted from a model can be used guide itself. Another type information training is knowledge provided by powerful teacher model. goal this work self-explanation borrowing ideas distillation. We start investigating effective components transferred network student network. Our investigation reveals that both responses in non-ground-truth classes and class-similarity teacher’s outputs contribute success Motivated conclusion, we propose an implementation introspective distilling online self-explanations. models trained with procedure outperform ones standard procedure, as well different regularization methods. When compared learned peer networks or networks, our also show competitive performance requires neither peers nor teachers.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Introspective Self-Explanation in Analytical Agents

There is a critical and urgent need for automated analytical agents operating in complex domains to provide meta-level explanations of their reasoning and conclusions. In this paper, we identify the principles for designing analytical agents that can explain their reasoning and justify their conclusions at different levels of abstractions to potential human customers with varying goals. We also...

متن کامل

Extending Introspective Learning from Self-Models

This position paper presents open issues for using self-models to guide introspective learning, focusing on five key types of areas to explore: (1) broadening the range of learning focuses and the range of learning tools which may be brought to bear, (2) learning for self-understanding as well as self-repair, (3) making model-based approaches more sensitive to processing characteristics, instea...

متن کامل

Face Model Compression by Distilling Knowledge from Neurons

The recent advanced face recognition systems were built on large Deep Neural Networks (DNNs) or their ensembles, which have millions of parameters. However, the expensive computation of DNNs make their deployment difficult on mobile and embedded devices. This work addresses model compression for face recognition, where the learned knowledge of a large teacher network or its ensemble is utilized...

متن کامل

Distilling Task Knowledge from How-To Communities

Knowledge graphs have become a fundamental asset for search engines. A fair amount of user queries seek information on problem-solving tasks such as building a fence or repairing a bicycle. However, knowledge graphs completely lack this kind of how-to knowledge. This paper presents a method for automatically constructing a formal knowledge base on tasks and task-solving steps, by tapping the co...

متن کامل

Distilling Model Knowledge

Top-performing machine learning systems, such as deep neural networks, large ensembles and complex probabilistic graphical models, can be expensive to store, slow to evaluate and hard to integrate into larger systems. Ideally, we would like to replace such cumbersome models with simpler models that perform equally well. In this thesis, we study knowledge distillation, the idea of extracting the...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2021

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-030-69538-5_3